Manipulation
Reversible, detachable robotic hand redefines dexterity
With its opposable thumb, multiple joints and gripping skin, human hands are often considered to be the pinnacle of dexterity, and many robotic hands are designed in their image. But having been shaped by the slow process of evolution, human hands are far from optimized, with the biggest drawbacks including our single, asymmetrical thumbs and attachment to arms with limited mobility. "We can easily see the limitations of the human hand when attempting to reach objects underneath furniture or behind shelves, or performing simultaneous tasks like holding a bottle while picking up a chip can," says Aude Billard, head of the Learning Algorithms and Systems Laboratory (LASA) in EPFL's School of Engineering. "Likewise, accessing objects positioned behind the hand while keeping the grip stable can be extremely challenging, requiring awkward wrist contortions or body repositioning." A team composed of Billard, LASA researcher Xiao Gao, and Kai Junge and Josie Hughes from the Computational Robot Design and Fabrication Lab designed a robotic hand that overcomes these challenges.
- Information Technology > Artificial Intelligence > Robots > Manipulation (0.97)
- Information Technology > Artificial Intelligence > Robots > Robots in the Workplace (0.87)
Appendix 367 A Implementation Details
W e are also committed to releasing the code. Implementation details for Stage 2. Our implementation strictly follows the previous work that also In this section, we briefly introduce our tasks. It requires the robot hand to open the door on the table. It requires the robot hand to orient the pen to the target orientation. It requires the robot hand to place the object on the table into the mug. We present the success rates of our six task categories as in Table 1.
Vine-inspired robotic gripper gently lifts heavy and fragile objects
In the horticultural world, some vines are especially grabby. As they grow, the woody tendrils can wrap around obstacles with enough force to pull down entire fences and trees. Inspired by vines' twisty tenacity, engineers at MIT and Stanford University have developed a robotic gripper that can snake around and lift a variety of objects, including a glass vase and a watermelon, offering a gentler approach compared to conventional gripper designs. A larger version of the robo-tendrils can also safely lift a human out of bed. The new bot consists of a pressurized box, positioned near the target object, from which long, vine-like tubes inflate and grow, like socks being turned inside out.
- North America > United States > Texas (0.05)
- North America > United States > Florida > Alachua County > Gainesville (0.05)
- Health & Medicine (0.49)
- Leisure & Entertainment > Sports > Soccer (0.30)
This robot hand can detach from its arm and crawl around
Breakthroughs, discoveries, and DIY tips sent six days a week. Engineers in Switzerland recently created a detachable, spider-like robot hand capable of grabbing multiple objects and using its fingers to crawl. The unsettling device, reminiscent of a threatening video game creature, can separate itself from a mounted robot arm, tip-toe (or really, tip-) its way toward small objects, pick them up, and carry them on its back. The symmetrical design and flexible fingers mean that the robot can transport objects on either side of its body. For humans, that would look like holding a ball in your palm while simultaneously grasping a piece of fruit on the back of your hand.
- North America > United States > New York (0.05)
- Europe > Switzerland > Vaud > Lausanne (0.05)
Sharpa's ping-pong playing, blackjack dealing humanoid is working overtime at CES 2026
Sharpa's ping-pong playing, blackjack dealing humanoid is working overtime at CES 2026 The company's super dexterous robotic hand, SharpaWave, allows it to pull individual playing cards from a deck. There were no idle hands at Sharpa's CES booth. The company's humanoid may have been the busiest bot at show, autonomously playing ping-pong, dealing blackjack games and taking selfies with passersby. The hand has 22 active degrees of freedom, according to the company, allowing for precise and intricate finger movements. It mirrored my gestures as I wiggled my hand in front of its camera, getting everything mostly right, which was honestly pretty cool.
Robohub highlights 2025
Over the course of the year, we've had the pleasure of working with many talented researchers from across the globe. As 2025 draws to a close, we take a look back at some of the excellent blog posts, interviews and podcasts from our contributors. Jiahui Zhang and Jesse Zhang to tell us about their framework for learning robot manipulation tasks solely from language instructions without per-task demonstrations. Hui Zhang writes about work presented at CoRL2025 on RobustDexGrasp, a novel framework that tackles different grasping challenges with targeted solutions. In this podcast from AAAI, host Ella Lan asked Professor Marynel Vázquez about what inspired her research direction, how her perspective on human-robot interactions has changed over time, robots navigating the social world, and more.
- South America > Brazil (0.05)
- North America > United States > Oregon (0.05)
- North America > United States > Ohio (0.05)
- (2 more...)
- Instructional Material > Course Syllabus & Notes (0.35)
- Personal > Interview (0.34)
- Information Technology > Communications (1.00)
- Information Technology > Artificial Intelligence > Robots > Manipulation (0.35)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.32)
- Asia > Japan > Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Information Technology (0.46)
- Education (0.46)
MaskedManipulator: Versatile Whole-Body Manipulation
Tessler, Chen, Jiang, Yifeng, Coumans, Erwin, Luo, Zhengyi, Chechik, Gal, Peng, Xue Bin
We tackle the challenges of synthesizing versatile, physically simulated human motions for full-body object manipulation. Unlike prior methods that are focused on detailed motion tracking, trajectory following, or teleoperation, our framework enables users to specify versatile high-level objectives such as target object poses or body poses. To achieve this, we introduce MaskedManipulator, a generative control policy distilled from a tracking controller trained on large-scale human motion capture data. This two-stage learning process allows the system to perform complex interaction behaviors, while providing intuitive user control over both character and object motions. MaskedManipulator produces goal-directed manipulation behaviors that expand the scope of interactive animation systems beyond task-specific solutions.
- Asia > China > Hong Kong (0.04)
- North America > Canada (0.04)
- Asia > Middle East > Israel (0.04)
- (3 more...)
ShapeForce: Low-Cost Soft Robotic Wrist for Contact-Rich Manipulation
Zhu, Jinxuan, Yan, Zihao, Xiao, Yangyu, Guo, Jingxiang, Tie, Chenrui, Cao, Xinyi, Zheng, Yuhang, Shao, Lin
Contact feedback is essential for contact-rich robotic manipulation, as it allows the robot to detect subtle interaction changes and adjust its actions accordingly. Six-axis force-torque sensors are commonly used to obtain contact feedback, but their high cost and fragility have discouraged many researchers from adopting them in contact-rich tasks. To offer a more cost-efficient and easy-accessible source of contact feedback, we present ShapeForce, a low-cost, plug-and-play soft wrist that provides force-like signals for contact-rich robotic manipulation. Inspired by how humans rely on relative force changes in contact rather than precise force magnitudes, ShapeForce converts external force and torque into measurable deformations of its compliant core, which are then estimated via marker-based pose tracking and converted into force-like signals. Our design eliminates the need for calibration or specialized electronics to obtain exact values, and instead focuses on capturing force and torque changes sufficient for enabling contact-rich manipulation. Extensive experiments across diverse contact-rich tasks and manipulation policies demonstrate that ShapeForce delivers performance comparable to six-axis force-torque sensors at an extremely low cost.
- Asia > Singapore > Central Region > Singapore (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
Simultaneous Tactile-Visual Perception for Learning Multimodal Robot Manipulation
Li, Yuyang, Chen, Yinghan, Zhao, Zihang, Li, Puhao, Liu, Tengyu, Huang, Siyuan, Zhu, Yixin
Robotic manipulation requires both rich multimodal perception and effective learning frameworks to handle complex real-world tasks. See-through-skin (STS) sensors, which combine tactile and visual perception, offer promising sensing capabilities, while modern imitation learning provides powerful tools for policy acquisition. However, existing STS designs lack simultaneous multimodal perception and suffer from unreliable tactile tracking. Furthermore, integrating these rich multimodal signals into learning-based manipulation pipelines remains an open challenge. We introduce TacThru, an STS sensor enabling simultaneous visual perception and robust tactile signal extraction, and TacThru-UMI, an imitation learning framework that leverages these multimodal signals for manipulation. Our sensor features a fully transparent elastomer, persistent illumination, novel keyline markers, and efficient tracking, while our learning system integrates these signals through a Transformer-based Diffusion Policy. Experiments on five challenging real-world tasks show that TacThru-UMI achieves an average success rate of 85.5%, significantly outperforming the baselines of alternating tactile-visual (66.3%) and vision-only (55.4%). The system excels in critical scenarios, including contact detection with thin and soft objects and precision manipulation requiring multimodal coordination. This work demonstrates that combining simultaneous multimodal perception with modern learning frameworks enables more precise, adaptable robotic manipulation.
- Asia > China > Hubei Province > Wuhan (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Information Technology > Artificial Intelligence > Robots > Manipulation (0.84)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.34)